-
Notifications
You must be signed in to change notification settings - Fork 1.7k
Issues: mlc-ai/mlc-llm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Feature Request] Implement quantized KV cache, storing it in 8-bit precision.
feature request
New feature or request
#3236
opened May 29, 2025 by
olyasir
[Question] TVMError: Binary was created using {relax.Executable} but a loader of that name is not registered.
question
Question about the usage
#3234
opened May 25, 2025 by
lixiaolong0424
[Question] Is there support for EXAONE-3.5 model from LG AI
question
Question about the usage
#3232
opened May 20, 2025 by
mansur20478
[Bug] CUDA cudaErrorSymbolNotFound during inference with Qwen3-30B-A3B-q4f16_1-MLC model on Tesla P100
bug
Confirmed bugs
#3231
opened May 15, 2025 by
rankaiyx
[Question] ValueError: invalid literal for int() with base 10: '8_e5m2'
question
Question about the usage
#3229
opened May 14, 2025 by
wzdsnd
[Question] How to run quickstart Chat Completion script?
question
Question about the usage
#3227
opened May 12, 2025 by
baqwas
[Bug] KeyError: 'qwen3' when running mlc_llm package
bug
Confirmed bugs
#3226
opened May 11, 2025 by
sandoche
[Question] ValueError about Invoking function image_embed in llava-1.5-7b-hf
question
Question about the usage
#3225
opened May 9, 2025 by
nathaniel2020
[Question] Does mlc-llm support HarmonyOS?
question
Question about the usage
#3217
opened Apr 30, 2025 by
Stockton11
[Bug] Cannot auto device detect without internet
bug
Confirmed bugs
#3214
opened Apr 26, 2025 by
Raviu56
[Bug] Speculative decoding not working due to difference in vocab_size (Qwen2.5 serie)
bug
Confirmed bugs
#3212
opened Apr 22, 2025 by
glennhanks
[Question] Should we expect generation quality comparable to gguf with 4-bit quantization?
question
Question about the usage
#3210
opened Apr 18, 2025 by
bene-ges
Failed mlc build from source with CUTLASS enabled
bug
Confirmed bugs
#3208
opened Apr 18, 2025 by
notabd7-deepshard
[Model Request] BitNet b1.58 2B4T - Scaling Native 1-bit LLM
new-models
#3207
opened Apr 17, 2025 by
pikann
[Bug] Trouble to run Confirmed bugs
mlc_llm chat
with Gemma 3 models
bug
#3206
opened Apr 16, 2025 by
grf53
[Bug] Missing post layernorm in CLIP model
bug
Confirmed bugs
#3205
opened Apr 16, 2025 by
vincentccc
[Bug] CMake Error at 3rdparty/tokenizers-cpp/msgpack/CMakeLists.txt during CMake iOS
bug
Confirmed bugs
#3204
opened Apr 15, 2025 by
KingSlayer06
Previous Next
ProTip!
Updated in the last three days: updated:>2025-05-31.